Introduction
Fútbol has been an integral part of my journey from the first time I stepped onto the field. The thrill of playing as a left winger, the strategy behind every move, and the joy of scoring have deeply influenced my understanding and love for the sport. This passion has seamlessly intertwined with my growing interest in analytics, leading me to explore the untapped potential of data in revealing hidden aspects of the game, such as player dynamics, strategic patterns, and performance metrics.
Liga MX is more than just a league to me; it’s a big part of Mexican culture. The Apertura and Clausura tournaments show the excitement and spirit of Mexican soccer, bringing stories of victory and defeat that engage fans every year.
In this study, I’ll concentrate on the key players in defense: the center-backs. Their role, often overshadowed by the flamboyance of attackers, is crucial in orchestrating the team’s defensive harmony. Leveraging unsupervised machine learning techniques such as Principal Component Analysis (PCA), K-means clustering, and Hierarchical Clustering, I aim to dissect and categorize the playing styles of these pivotal figures. Each technique serves a purpose: PCA to reduce dimensionality and highlight important performance indicators, K-means to group similar player profiles based on these indicators, and Hierarchical Clustering to build a detailed taxonomy of player types based on their on-field behavior.
This study is more than just research; it’s an exploration into Liga MX using data analytics. By applying these unsupervised learning methods, we will uncover the nuanced contributions of center-backs, offering new insights and appreciation for their role in the beautiful game. Join me in this analytical journey, where numbers meet the pitch, and discover the unseen stories of Liga MX’s defensive performances.
Data
For this research, I utilized a package tool called worldfootballr
, which is designed to pull global football (soccer) data from platforms like FBref, Transfermarkt, Understat, and FotMob. This tool helped me gather advanced statistics provided by Opta, which were then formatted into a CSV file capturing player performances throughout the Clausura 2024 regular season. Initially, the dataset comprised 2,428 entries across 47 different columns. In the process of refining this dataset, I removed information irrelevant to our specific analysis. This process also involved using a histogram shown in Figure 1 of each player’s total minutes to distinguish those who had accumulated at least 120 minutes of playtime from the start of the season until matchday 9.
After these adjustments, the final dataset presented 57 unique center-back players alongside 15 variable features. Each record within this dataset corresponds to an individual player, encompassing important performance data from the season, such as the number of tackles made, tackles within the defensive third, and instances of lost challenges. Below in Table 1 are the first 6 observations:
Tkl_Tackles | TklW_Tackles | Def 3rd_Tackles | Mid 3rd_Tackles | Att 3rd_Tackles | Tkl_Challenges | Att_Challenges | Lost_Challenges | Blocks_Blocks | Sh_Blocks | Pass_Blocks | Int | Tkl+Int | Clr | Err |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
2.11 | 1.22 | 1.33 | 0.78 | 0 | 1.00 | 1.44 | 0.44 | 1.11 | 0.89 | 0.22 | 1.44 | 3.56 | 4.00 | 0 |
1.44 | 1.11 | 1.11 | 0.33 | 0 | 1.22 | 1.33 | 0.11 | 1.22 | 0.33 | 0.89 | 0.89 | 2.33 | 6.78 | 0 |
2.11 | 1.33 | 1.67 | 0.44 | 0 | 1.22 | 1.78 | 0.56 | 1.56 | 1.00 | 0.56 | 2.22 | 4.33 | 6.00 | 0 |
2.00 | 1.29 | 1.86 | 0.14 | 0 | 0.71 | 1.00 | 0.29 | 1.43 | 0.71 | 0.71 | 0.86 | 2.86 | 3.57 | 0 |
2.33 | 1.00 | 1.67 | 0.67 | 0 | 2.00 | 3.33 | 1.33 | 0.33 | 0.33 | 0.00 | 2.67 | 5.00 | 1.67 | 0 |
0.33 | 0.22 | 0.22 | 0.11 | 0 | 0.11 | 0.11 | 0.00 | 0.56 | 0.44 | 0.11 | 0.56 | 0.89 | 2.89 | 0 |
Below in Table 2 are the description of each variable used for the analysis:
Variables |
Description |
---|---|
Tkl_Tackles | Number of players tackled |
TklW_Tackles | Tackles in which the tackler's team won possession of the ball |
Def 3rd_Tackles | Tackles in defensive 1/3 |
Mid 3rd_Tackles | Tackles in middle 1/3 |
Att 3rd_Tackles | Tackles in attacking 1/3 |
Tkl_Challenges | Number of dribblers tackled |
Att_Challenges | Number of unsuccessful challenges plus number of dribblers tackled |
Lost_Challenges | Number of unsucessful attempts to challenge a dribbling player |
Blocks_Blocks | Number of times blocking the ball by standing in its path |
Sh_Blocks | Number of times blocking a shot by standing in its path |
Pass_Blocks | Number of times blocking a pass by standing in its path |
Int | Interceptions |
Tkl+Int | Number of players tackled plus number of interceptions |
Clr | Clearances |
Err | Mistakes leading to an opponent's shot |
Principal Component Analysis (PCA)
Principal component analysis aims to describe the variance-covariance relationship among a group of variables by using a small number of linear combinations of these variables. The primary goals are to reduce the dimensionality of the data and to provide an interpretation through the rotation process.
Algebraically, principal components are unique linear combinations of the p random variables \(X_1,X_2,\ldots,X_p\). Geometrically, this translates to forming a new set of axes through the rotation of the original coordinate system, where \(X_1,X_2,\ldots,X_p\) are the axes. These newly established axes stand for the most significant variation directions, simplifying and condensing the description of the covariance matrix.
Let the random vector \(X' = [X_1,X_2,\dots,X_p]\) have the covariance matrix \(\Sigma\) with eigenvalues \(\lambda_1 \geq \lambda_2 \geq \ldots \geq \lambda_p = 0\).
Consider the linear combinations \[\begin{align*} Y_1 &= \boldsymbol{a'_1X} = a_{11}X_1 + a_{12}X_2 +\ldots+ a_{1p}X_p \\ Y_2 &= \boldsymbol{a'_2X} = a_{21}X_1 + a_{22}X_2 +\ldots+ a_{2p}X_p \\ &\vdots\\ Y_p &= \boldsymbol{a'_pX} = a_{p1}X_1 + a_{p2}X_2 +\ldots+ a_{pp}X_p \end{align*}\]
Then, we obtain \[\begin{align*} \text{Var}(Y_i) &= \boldsymbol{a'_i\Sigma a_i} \qquad &i = 1,2,...,p \\ \text{Cov}(Y_i,Y_k) &= \boldsymbol{a'_i\Sigma a_k} \qquad &i,k = 1,2,\ldots,p \end{align*}\]
The principal components are distinct linear combinations formed from the original variables, designed to be independent of each other.
The first principal component is the linear combination with maximum variance. That is, it maximizes \(\text{Var}(Y_1) = \boldsymbol{a'_i\Sigma a_i}\). It is clear that \(\text{Var}(Y_1) = \boldsymbol{a'_i\Sigma a_i}\) can be increased by multiplying any \(\boldsymbol{a_1}\) by some constant. To eliminate this indeterminacy, it is convenient to restrict attention to coefficient vectors of unit length. We therefore define
\[\begin{align*} \text{First principal component} =\ &\text{linear combination}\ \boldsymbol{a'_1X}\ \text{that maximizes} \\ &\text{Var}(\boldsymbol{a'_1X})\ \text{subject to a}\ \boldsymbol{a'_1a_1} = 1 \end{align*}\] \[\begin{align*} \text{Second principal component} =\ &\text{linear combination}\ \boldsymbol{a'_2X}\ \text{that maximizes} \\ &\text{Var}(\boldsymbol{a'_2X})\ \text{subject to a}\ \boldsymbol{a'_2a_2} = 1\ \text{and} \\ &\text{Cov}(\boldsymbol{a'_1X,a'_2X}) = 0 \end{align*}\]At the \(i\)th step, \[\begin{align*} i\text{th principal component} =\ &\text{linear combination}\ \boldsymbol{a'_iX}\ \text{that maximizes} \\ &\text{Var}(\boldsymbol{a'_iX})\ \text{subject to a}\ \boldsymbol{a'_ia_i} = 1\ \text{and} \\ &\text{Cov}(\boldsymbol{a'_iX,a'_kX}) = 0\ \text{for}\ k < i \end{align*}\]
Initially, the data is scaled to normalize the scale of each variable. This ensures that all variables contribute equally to the analysis. This standardization process is crucial and is mathematically represented as,
\[\begin{align*} Z_1 &= \frac{(X_1 - \mu_1)}{\sqrt{\sigma_{11}}} \\ Z_2 &= \frac{(X_2 - \mu_2)}{\sqrt{\sigma_{22}}} \\ &\vdots\\ Z_p &= \frac{(X_p - \mu_p)}{\sqrt{\sigma_{pp}}} \end{align*}\]In matrix form, \[\begin{align*} \boldsymbol{Z} = (\boldsymbol{V}^{1/2})^{-1}(\boldsymbol{X} - \boldsymbol{\mu}) \end{align*}\]
where the diagonal standard deviation matrix is \(\boldsymbol{V}^{1/2}\).
After standardizing the data, we compute its covariance matrix to examine the relationships among the variables, which is mathematically represented as:
\[\begin{align*} \text{Cov}(\boldsymbol{Z}) = (\boldsymbol{V}^{1/2})^{-1}\boldsymbol{\Sigma}(\boldsymbol{V}^{1/2})^{-1} = \boldsymbol{\rho} \end{align*}\]In this formula, \(\boldsymbol{\rho}\) denotes the correlation matrix of the original variables \(\boldsymbol{X}\), after transformation into the standardized form \(\boldsymbol{Z}\). This step is crucial for understanding how variables relate to one another in the scaled space. Following this, we determine the eigenvalues and eigenvectors from this correlation matrix, \(\boldsymbol{\rho}\). The eigenvalues quantify the amount of variance each principal component captures, while the eigenvectors specify the directional orientation of these components, providing insight into the underlying structure of the data.
The standardized data is shown below in Table 3:
Tkl_Tackles | TklW_Tackles | Def 3rd_Tackles | Mid 3rd_Tackles | Att 3rd_Tackles | Tkl_Challenges | Att_Challenges | Lost_Challenges | Blocks_Blocks | Sh_Blocks | Pass_Blocks | Int | Tkl+Int | Clr | Err | |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Adonis Frías | 1.08 | 0.77 | 0.74 | 1.39 | -0.55 | 0.69 | 0.44 | 0.03 | 0.14 | 0.57 | -0.53 | 0.34 | 0.90 | 0.42 | -0.34 |
Alexis Francisco Peña | 0.08 | 0.53 | 0.34 | -0.36 | -0.55 | 1.19 | 0.29 | -0.84 | 0.34 | -0.73 | 1.54 | -0.54 | -0.27 | 2.48 | -0.34 |
Alán Montes | 1.08 | 1.01 | 1.36 | 0.06 | -0.55 | 1.19 | 0.92 | 0.35 | 0.97 | 0.82 | 0.52 | 1.59 | 1.63 | 1.90 | -0.34 |
Anderson Santamaría | 0.92 | 0.92 | 1.71 | -1.10 | -0.55 | 0.03 | -0.17 | -0.36 | 0.73 | 0.15 | 0.98 | -0.59 | 0.23 | 0.10 | -0.34 |
Andres Micolta Arroyo | 1.41 | 0.29 | 1.36 | 0.96 | -0.55 | 2.97 | 3.08 | 2.39 | -1.31 | -0.73 | -1.21 | 2.32 | 2.26 | -1.30 | -0.34 |
Antonio Briseño | -1.57 | -1.41 | -1.27 | -1.22 | -0.55 | -1.33 | -1.41 | -1.13 | -0.88 | -0.47 | -0.87 | -1.08 | -1.63 | -0.40 | -0.34 |
In our analysis, we specifically focus on the loadings of the first two principal components. These loadings reflect how each original variable contributes to the formation of the principal components. By analyzing the loadings of the first two components, we gain insights into the most significant patterns and variations within the data.
The mathematical approach of the \(i\)th principal component of the standardized variables \(\boldsymbol{Z}' = [Z_1,Z_2,\ldots,Z_p]\) with \(\text{Cov}(\boldsymbol{Z}) = \boldsymbol{\rho}\) is represented as:
\[\begin{align*} \boldsymbol{Y_i}' = \boldsymbol{{e_i{Z}}} = \boldsymbol{e'_{i}}(\boldsymbol{V}^{1/2})^{-1}(\boldsymbol{X}-\boldsymbol{\mu}), \qquad i = 1,2,\ldots,p \end{align*}\]The loadings for the first and second principal components, denoted as PC1 and PC2 respectively, are presented below in Table 4:
Code
<- cov(scaled_df)
cov_df <- eigen(cov_df)
eigen_df
<- eigen_df$vectors[,1:2]
phi
row.names(phi) <- colnames(clean_df)
colnames(phi) <- c("PC 1", "PC 2")
%>%
phi ::kable(digits = 2) knitr
PC 1 | PC 2 | |
---|---|---|
Tkl_Tackles | -0.37 | -0.12 |
TklW_Tackles | -0.33 | -0.07 |
Def 3rd_Tackles | -0.34 | -0.06 |
Mid 3rd_Tackles | -0.23 | -0.13 |
Att 3rd_Tackles | 0.01 | -0.27 |
Tkl_Challenges | -0.33 | -0.27 |
Att_Challenges | -0.33 | -0.20 |
Lost_Challenges | -0.26 | -0.07 |
Blocks_Blocks | -0.20 | 0.55 |
Sh_Blocks | -0.11 | 0.54 |
Pass_Blocks | -0.18 | 0.19 |
Int | -0.20 | 0.10 |
Tkl+Int | -0.36 | -0.02 |
Clr | -0.19 | 0.35 |
Err | -0.08 | 0.06 |
The total variance for the standardized variables in the population is equal to \(p\), which represents the sum of the diagonal elements of the matrix, also known as the trace of the correlation matrix \(\boldsymbol{\rho}\). Consequently, the fraction of the total variance accounted for by the \(k\)th principal component of \(\boldsymbol{Z}\) is determined as follows:
\[ \begin{pmatrix} \text{Proportion of (standardized)}\\ \text{population variance due}\\ \text{to}\ k\text{th principal component} \end{pmatrix} = \frac{\lambda_k}{p}, \qquad k = 1,2,\ldots,p \] where the \(\lambda_k's\) are the eigenvalues of \(\boldsymbol{\rho}\).
The Proportion of Variance for each Principal Component (PC) is presented below in Table 5:
Code
<- sum(diag(cov(scaled_df)))
sum_cov <- eigen_df$values/sum_cov
results
<- tibble(
prop_var principal_components = paste0("PC ", 1:15),
results = results
)
%>%
prop_var gt() %>%
cols_label(
principal_components = md("**Principal Components**"),
results = md("**Results**")
%>%
) fmt_number(
columns = results,
decimals = 2
%>%
) opt_table_font(
google_font(name = "Roboto")
%>%
) tab_options(data_row.padding = px(1))
Principal Components |
Results |
---|---|
PC 1 | 0.43 |
PC 2 | 0.13 |
PC 3 | 0.10 |
PC 4 | 0.08 |
PC 5 | 0.07 |
PC 6 | 0.07 |
PC 7 | 0.05 |
PC 8 | 0.03 |
PC 9 | 0.02 |
PC 10 | 0.02 |
PC 11 | 0.01 |
PC 12 | 0.00 |
PC 13 | 0.00 |
PC 14 | 0.00 |
PC 15 | 0.00 |
Now, the proportion of total variance explained by the first and second principal components is calculated as the sum of the first eigenvalue divided by the total variance plus the second eigenvalue divided by the total variance, represented as:
\[\begin{align*} (\frac{\lambda_{1} + \lambda_{2}}{p})100\% = (\frac{6.376 + 1.960}{15})100\% = 55.6\% \end{align*}\]of the total (standardized) variance, have interesting interpretations.
Code
<- (eigen_df$values[1] + eigen_df$values[2]) / sum_cov
total_prop_var
print(paste0("Total variance explained by PC1 and PC2: ", round(total_prop_var, 4) * 100, "%"))
[1] "Total variance explained by PC1 and PC2: 55.58%"
The PCA scores for each player, as shown below in Table 6, represent the coordinates of each player in the space defined by the first two principal components (PC1 and PC2). These scores are calculated based on the original data transformed by the respective principal component loadings.
Player | Team | PC1 | PC2 |
---|---|---|---|
Adonis Frías | León | −2.05 | −0.10 |
Alexis Francisco Peña | Necaxa | −1.03 | 0.69 |
Alán Montes | Necaxa | −3.63 | 1.18 |
Anderson Santamaría | Atlas | −1.13 | 0.69 |
Andres Micolta Arroyo | Pachuca | −4.36 | −3.45 |
Antonio Briseño | Guadalajara | 4.25 | 0.25 |
Scores For Each Player
Figure 2 indicates the presence of two outliers, specifically players Guido Pizarro from Tigres UANL and Andres Micolta Arroyo from CF Pachuca. We will examine the impact of these outliers on the clustering process in subsequent steps.
prcomp
Function
R offers an easy method for data analysis with its builtin prcomp
function, used for Principal Component Analysis. If you set scale = TRUE
, it standardizes the data automatically. This function provides principal components and their impact on variance. The summary given by the function shows detailed variance information for each component, making it easier to understand and reduce complex data.
Code
<- prcomp(clean_df, scale = TRUE)
pca_df
summary(pca_df)
Importance of components:
PC1 PC2 PC3 PC4 PC5 PC6 PC7
Standard deviation 2.5251 1.4001 1.19456 1.08645 1.04146 1.0188 0.84045
Proportion of Variance 0.4251 0.1307 0.09513 0.07869 0.07231 0.0692 0.04709
Cumulative Proportion 0.4251 0.5558 0.65090 0.72959 0.80190 0.8711 0.91819
PC8 PC9 PC10 PC11 PC12 PC13
Standard deviation 0.69325 0.57368 0.53408 0.36360 0.00566 0.005024
Proportion of Variance 0.03204 0.02194 0.01902 0.00881 0.00000 0.000000
Cumulative Proportion 0.95022 0.97217 0.99118 0.99999 1.00000 1.000000
PC14 PC15
Standard deviation 0.003521 0.002493
Proportion of Variance 0.000000 0.000000
Cumulative Proportion 1.000000 1.000000
The summary indicates that the initial principal components, particularly PC1 and PC2, play significant roles in capturing the dataset’s variance, with PC1 alone accounting for 42.51% and PC2 adding another 13.07%. This showcases their importance in representing the dataset’s structure. As more components are added, the cumulative variance explained increases, reaching virtually 100% by PC12, which suggests that the components beyond this point have minimal impact on explaining additional variance. That means that when \(\rho\) is near 1, the last \(\rho - 1\) components collectively contribute very little to the total variance and can often be neglected. Therefore, focusing on the first few components is essential for effective dimensionality reduction and dataset interpretation, while the latter components can generally be omitted from further analysis.
The scree plot in Figure 3 above is a useful visual aid to determining an appropriate number of principal components. The first principal component explain 42.51% of the total variance. The first two principal components, collectively, explain 55.58% of the total variance. An elbow occurs at about \(i = 3\). That is, the eigenvalues after \(\lambda_2\) are all relatively small and about the same size. In this case, it appears, without any other evidence, that two sample principal components effectively summarize the total variance of the data.
Contributions of variables to PCs
The key variables, which contribute most significantly, can be emphasized in the correlation plot as follows:
All the variables grouped together Figure 4, such as Tkl+Int
, Def_3rd_Tackles
, Tkl_Tackles
, Att_3rd_Tackles
, TklW_Tackles
, Tkl_Challenges
, Mid_3rd_Tackles
, Att_Challenges
, and Lost_Challenges
are positively correlated with each other. This positive correlation is due to their proximity to each other and their downward direction. Conversely, Sh_Blocks
, Blocks_Blocks
, and Clr
are grouped together and show a positive correlation, as their vectors point upwards. This indicates that these two sets of variables move in opposite directions, meaning they are negatively correlated. Negatively correlated variables are positioned on opposite sides of the plot’s origin, in opposed quadrants. The distance between variables and the origin measures how well the variables are represented on the factor map; those further from the origin are better represented.
This bar plot displays the contributions of various variables to PC1 (dim1) on the left and to PC2 (dim2) on the right. The red dashed line across the graph represents the expected average contribution for each dimension.
Biplot
In the biplot, individual player scores are plotted alongside variables, allowing for a comprehensive visual representation of their performances in relation to the different measured factors.
In this biplot analysis plot from Figure 5, it’s crucial to understand that a player’s position relative to a specific variable indicates their performance in that area. Players on the same side as a variable typically exhibit high values for that trait, while those on the opposite side represent lower values. For instance, several players located in the third quadrant, where most Dim1 contributing variables are present, demonstrate significant defensive actions. Sebastian Olmedo from Club Puebla, notable for his high Dim1 representation, exemplifies a player actively involved in defense. Similarly, players like Arturo Ortiz, who are positioned close to Olmedo, share these defensive characteristics, forming a group identified as traditional old-school defenders.
Conversely, in the second quadrant, opposite the traditional defenders, we find variables negatively correlated with defensive actions, such as Blocks_Blocks
, Sh_Blocks
, and Clr
. Santiago Nunez of Santos Laguna stands out here as a “stopper,” excelling in blocking, indicative of a different defensive skill set.
Additionally, outliers like Andres Micolta Arroyo from C.F. Pachuca and Guido Pizarro from Tigres UANL are positioned apart from clusters, suggesting unique styles of play that, while distinct, significantly impact the game by contributing to key variables.
Lastly, there’s a distinct group known as Att.3rd_tackles
players, including Diego de Buen from Club Puebla and Ventura Alvarado from Mazatlan F.C. These individuals specialize in stopping counter-attacks and play a crucial role during ‘el balón parado’ situations, as they call it in Spanish, which refers to set pieces or moments when the ball is stationary. Their positioning in the biplot underscores their expertise in initiating defensive maneuvers during these pivotal moments of the game.
Non-Hierarchical Clustering
K-Means Clustering
Following the PCA analysis, we will now proceed with K-means clustering to further explore our dataset. This method will allow us to categorize the data into distinct groups based on similarities identified during the PCA. By applying K-means clustering, we aim to enhance our understanding of the underlying structures and relationships within the data, facilitating more targeted interpretations and insights. This step is crucial in identifying natural groupings among players or variables, leading to more nuanced conclusions and strategic decisions.
A Focus on Euclidean Distance: An essential step in K-means clustering involves discussing the dissimilarity or distance measurement within the data. In our analysis, we opt for the Euclidean distance, which is the standard method for K-means and measures the straight-line distance between points in multidimensional space. This choice is due to its simplicity and effectiveness in quantifying the dissimilarity between data points. By utilizing Euclidean distance, we can ensure that the clusters formed are based on the actual spatial differences among data points, making it a suitable and intuitive approach for our clustering objectives.
The Euclidean distance between two p-dimensional points is given by the formula:
\[ d(x, y) = \sqrt{(x_1 - y_1)^2 + (x_2 - y_2)^2 + \cdots + (x_p - y_p)^2} = \sqrt{(x - y)'(x - y)} \]
K-Means Clustering: the process is composed of these three steps:
Partition the items into \(K\) initial clusters.
Proceed through the list of items, assigning an item to the cluster whose centroid (mean) is nearest. (Distance is usually computed using Euclidean distance with either standardized or standardized.) Recalculate the centroid for the cluster receiving the new item and for the cluster losing the item.
Repeat Step 2 until no more reassignments take place. In practice, such as in R, this process might begin with a default of 10 iterations to ensure stability and convergence of clusters.
To determine the optimal number of clusters (k) in our analysis, we employed two methods: the Within-Sum-of-Squares (WSS) and the Gap Statistic. The WSS method helps identify the elbow point where the addition of another cluster does not significantly improve the total variance within clusters. Meanwhile, the Gap Statistic compares the log-wss of our dataset against expected values under a null reference distribution of the data, providing a more robust estimation of the optimal k.
Based on the analysis plot from Figure 6 of the Total Within Sum of Square and the Gap Statistic plots, the optimal number of clusters for the data is recommended to be three. The Elbow Method suggests a bend at three or four clusters, while the Gap Statistic clearly peaks and maximizes at three clusters, indicating the most suitable partitioning of the data.
After identifying the optimal number of clusters as three through both the Elbow Method and the Gap Statistic, we proceed with the segmentation of the dataset to uncover distinct groupings and patterns. This allows us to visualize and analyze the inherent structures and relationships within our data.
- Light Green Cluster (Circle): “Defensores tradicionales”
Players in this cluster, like Sebastian Olmedo and Arturo Ortiz, exhibit strong defensive capabilities, as evidenced by the cluster’s alignment with key defensive actions such as tackles, interceptions, and blocks. This group is indicative of players who excel in defensive duties, demonstrating a high degree of proficiency in preventing opposition attacks. The outliers “Andres Micolta Arroyo” and “Guido Pizarro” may represent players with unique defensive styles or roles that differ from the standard defensive metrics, warranting closer examination.
- Red-Orange Cluster (Triangle): “Defendores diferentes”
This cluster features players such as Carlos Salcedo, who are typically positioned further to the right, suggesting a distinct playing style. These individuals may have fewer minutes on the pitch or adopt a defensive approach that doesn’t primarily focus on tackles, interceptions, and blocks. An example within this category is Hector Moreno of Monterrey, who exemplifies a style that balances traditional defensive duties with advanced playmaking and ball distribution skills. The outlier “Unai Bilbao” suggests a player who might either excel exceptionally in these areas or possess a unique set of skills differing from the main group.
- Purple Cluster (Square): “Defensores Versátiles”
This group, featuring Diego De Buen and Erik Lira, showcases versatility, balancing defensive responsibilities and some with the ability to support midfield and forward play, indicative of a modern, flexible center-back. This adaptability makes them valuable for strategies requiring defensive solidity and midfield support.
Agglomerative Hierarchical Clustering
Agglomerative Hierarchical Clustering is a popular method used in data analysis to group objects into clusters based on their similarities. Starting with each object in its own cluster, it iteratively combines the closest pairs into larger clusters until all objects are in a single group or a stopping criterion is met. This bottom-up approach produces a dendrogram, a tree-like diagram that illustrates the series of merges and the hierarchical structure of the clusters.
Applied the Ward.D2 approach, a specific type of Agglomerative Hierarchical Clustering, our main goal here is to reduce the sum of squared differences, also known as the ‘loss of information,’ across all clusters. This reduction relies on calculating the squared Euclidean distances among the points in our dataset, a method that aligns with Ward’s strategy to decrease the total variance within each cluster continuously. This technique results in clusters that are more evenly distributed and meaningful, facilitating a better interpretation of the dataset’s intrinsic structure since it generally forms clusters that are comparable in size.
The resulting dendrogram, illustrated below in Figure 8, visualizes the hierarchical structure and the process of cluster formation. It provides insights into the data’s underlying patterns and relationships by showing how individual observations are grouped into clusters based on their similarities and how these clusters are combined into larger clusters as we move up the hierarchy.
Observations:
The light green cluster at the bottom includes players such as Sebastian Olmedo and Arturo Ortiz, indicating a group likely characterized by traditional defensive skills.
The purple cluster at the top, with players like Jesús Orozco and Igor Lichnovsky, suggesting these may be more versatile defenders who contribute both defensively and in initiating play, possibly indicating a balance between defensive robustness and ability to support midfield or offensive actions.
The red-orange cluster in the middle features players like Hector Moreno and Diego Reyes, could represent those who are perhaps less involved in direct defensive actions maybe because they spend less time on the pitch, but might play roles in ball distribution or possess other unique attributes distinguishing them from the purely defensive or hybrid players.
Conclusion
It is clear that different players fulfill their essential roles as center-back defenders. The art of defending, often overshadowed by the allure of goals and assists, remains the cornerstone of every successful football team. No team can dominate their league or excel in individual competitions without a strong defensive line. Employing machine learning classification methods, including PCA, non-hierarchical k-means, and hierarchical clustering, holds profound significance in sports analytics, particularly in identifying similarities among players based on their defensive traits. This approach can significantly aid teams in making informed decisions, especially during the transfer window, by identifying the ideal replacements that fit a specific defensive role, thereby enhancing team performance.
However, it is crucial to acknowledge significant limitations. For instance, there are skilled center-backs returning from injuries who have played minimally this season, which may influence their current performance levels as they regain their form. Additionally, some teams may experience a poor start to the season but later find themselves on a winning streak. The distinction between home and away games can also significantly impact player performance, with the venue potentially playing a crucial role in each match’s outcome. Furthermore, matches against top-tier teams in the tournament can demand more from a player defensively, contrasting with games against weaker teams where the defensive workload might be substantially less.
Future enhancements for this project should consider several factors. Initially, the data was collected from the beginning of the tournament, from matchday 1 to matchday 9, which only represents the mid-point of the season. Since each season (Apertura and Clausura) comprises 17 matchdays, a more precise model could be developed with information spanning the entire season. Additionally, incorporating data on full-backs, who also play defensive roles, could alter the classification narrative of players significantly. Lastly, enriching the dataset with additional metrics such as possession and aerial duels could further refine the differentiation of player types in defensive positioning.
Acknowledgments
I would like to express my sincere gratitude to the faculty members who guided me through my graduate studies in Applied Statistics at the Department of Mathematics and Statistics, California State University - Long Beach. Special acknowledgments to Dr. Sung Kim, Dr. Kagba Suaray, Dr. Olga Korosteleva, Dr. Rebecca Le, and Dr. Hojin Moon for their invaluable mentorship and support during my academic journey.
References
Johnson, R. A., & Wichern, D. W. (2007). Applied Multivariate Statistical Analysis. Prentice Hall.
Nadir, N. (2021, September 21). Assessing NBA Player Similarity with Machine Learning. Medium, Towards Data Science. https://towardsdatascience.com/which-nba-players-are-most-similar-machine-learning-provides-the-answers-r-project-b903f9b2fe1f.
OUKHOUYA, H. (2023, April 2). NBA Player Performance Analysis: PCA, Hierarchical Clustering, and K-Means Clustering. RPubs. https://rpubs.com/HassanOUKHOUYA/NBA_Player_Performance_Analysis_PCA_Hierarchical_Clustering_and_K_Means_Clustering.